Bernstein inequalities (probability theory)

In probability theory, Bernstein inequalities give bounds on the probability that the sum of random variables deviates from its mean. In the simplest case, let X1, ..., Xn be independent Bernoulli random variables taking values +1 and −1 with probability 1/2, then for every positive \varepsilon,

 \mathbf{P} \left\{\left|\;\frac{1}{n}\sum_{i=1}^n X_i\;\right| > \varepsilon \right\} \leq 2\exp \left\{ - \frac{n\varepsilon^2}{ 2 (1 %2B \varepsilon/3) } \right\}.

Bernstein inequalities were proved and published by Sergei Bernstein in the 1920s and 1930s.[1][2][3][4] Later, these inequalities were rediscovered several times in various forms. Thus, special cases of the Bernstein inequalities are also known as the Chernoff bound, Hoeffding's inequality and Azuma's inequality.

Contents

Some of the inequalities

1. Let X1, ..., Xn be independent zero-mean random variables. Suppose that |X i| ≤ M almost surely, for all i. Then, for all positive t,

\mathbf{P} \left\{ \sum_{i=1}^n X_i > t \right\} \leq \exp \left\{ - \frac{t^2/2}{\sum \mathbf{E} X_j^2 %2B Mt/3 } \right\}.

2. Let X1, ..., Xn be independent random variables. Suppose that for some positive real L and every integer k > 1,

 \mathbf{E} |X_i^k| \leq \frac{\mathbf{E} X_i^2}{2} L^{k-2} k!

Then

 \mathbf{P} \left\{ \sum_{i=1}^n X_i \geq 2 t \sqrt{\sum \mathbf{E} X_i^2} \right\}
   < \exp \left\{ - t^2\right\}, \text{ for } 0 < t \leq \frac{\sqrt{\sum \mathbf{E} X_j^2}}{2L}.

3. Let X1, ..., Xn be independent random variables. Suppose that

 \mathbf{E} |X_i^k| \leq \frac{k!}{4!} \left(\frac{L}{5}\right)^{k-4}

for all integer k > 3. Denote  A_k = \sum \mathbf{E} X_i^k . Then,

 \mathbf{P} \left\{ \left| \sum_{j=1}^n X_j - \frac{A_3 t^2}{3A_2} \right| 
   \geq \sqrt{2A_2} \, t \left[ 1 %2B \frac{A_4 t^2}{6 A_2^2} \right] \right\}
   < 2 \exp \left\{ - t^2\right\},\text{ for } 0 < t \leq \frac{5 \sqrt{2A_2}}{4L}.

4. Bernstein also proved generalizations of the inequalities above to weakly dependent random variables. For example, inequality (2) can be extended as follows. Let X1, ..., Xn be possibly non-independent random variables. Suppose that for all integer i > 0,

  1. \mathbf{E} \left\{ X_{i} | X_1, \dots, X_{i-1} \right\} = 0,
  2.  \mathbf{E} \left\{ X_i^2 | X_1, \dots, X_{i-1} \right\} \leq R_i\; \mathbf{E} X_i^2,
  3. \mathbf{E} \left\{ X_i^k | X_1, \dots, X_{i-1} \right\} 
\leq  \frac{\mathbf{E} \left\{ X_i^2 | X_1, \dots, X_{i-1} \right\}}{2} \; L^{k-2} k!

Then

 \mathbf{P} \left\{ \sum_{i=1}^n X_i \geq 2 t \sqrt{\sum_{i=1}^n R_i \mathbf{E} X_i^2} \right\} 
< \exp(-t^2), \text{ for } 0 < t \leq \frac{\sqrt{\sum_{i=1}^n R_i \mathbf{E} X_i^2}}{2L}.

Proofs

The proofs are based on an application of Markov's inequality to the random variable  \exp \left\{ \lambda \sum_{j=1}^n X_j \right\} , for a suitable choice of the parameter  \lambda > 0 .

See also

References

(according to: S.N.Bernstein, Collected Works, Nauka, 1964)

  1. ^ S.N.Bernstein, "On a modification of Chebyshev’s inequality and of the error formula of Laplace" vol. 4, #5 (original publication: Ann. Sci. Inst. Sav. Ukraine, Sect. Math. 1, 1924)
  2. ^ Bernstein, S. N. (1937). "[On certain modifications of Chebyshev's inequality]". Doklady Akademii Nauk SSSR 17 (6): 275–277. 
  3. ^ S.N.Bernstein, "Theory of Probability" (Russian), Moscow, 1927
  4. ^ J.V.Uspensky, "Introduction to Mathematical Probability", McGraw-Hill Book Company, 1937

A modern translation of some of these results can also be found in M. Hazewinkel (trans.), Encyclopedia of Mathematics, Kluwer, 1987, under the entry "Bernstein inequality" (online version).